BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT
:
Pre-training
of
Deep Bidirectional Transformers
for
Language Understanding
Jacob
Devlin
,
Ming-Wei Chang
,
Kenton Lee
, Kristina
Toutanova
Google AI Language
Submitted on 11 Oct 2018 (v1), last revised 24 May 2019 (this version, v2)
https://arxiv.org/abs/1810.04805